Enterprise Database Systems
Apache Kafka
Apache Kafka Development
Apache Kafka Operations
Clustering
Kafka Integration with Spark
Kafka Integration with Storm
Real-time Applications

Apache Kafka Development

Course Number:
df_apka_a02_it_enus
Lesson Objectives

Apache Kafka Development

  • start the course
  • describe the high-level consumer API for reading from Apache Kafka
  • describe the simple consumer API for reading from Apache Kafka
  • describe the Hadoop consumer API for reading from Apache Kafka
  • configure Apache Kafka brokers
  • configure Apache Kafka consumers
  • configure Apache Kafka producers
  • configure compression in Apache Kafka
  • describe the producer API in Apache Kafka
  • describe the SyncProducer API in Apache Kafka
  • describe the AsyncProducer API in Apache Kafka
  • configure message acknowledgement, or acking, in Apache Kafka
  • batch messages in Apache Kafka
  • specify keyed and non-keyed messages in Apache Kafka
  • configure broker discovery in Apache Kafka
  • use Apache Kafka test suites for testing
  • configure serialization and deserialization in Apache Kafka
  • build a custom serializer in Apache Kafka
  • configure a broker and create a producer and a consumer in Apache Kafka

Overview/Description
Apache Kafka comes with a set of APIs for consumers and producers for writing to and reading from logs. This course covers the producer and consumer APIs, and data serialization and deserialization techniques, and strategies for testing Kafka.

Target Audience
Developers, IT Operations engineers, and DevOPs engineers looking to implement and manage Apache Kafka

Apache Kafka Operations

Course Number:
df_apka_a01_it_enus
Lesson Objectives

Apache Kafka Operations

  • start the course
  • describe the function of Apache Kafka
  • describe the architecture of Apache Kafka
  • describe Apache Kafka topics
  • describe Apache Kafka partitions
  • describe Apache Kafka replicas
  • describe Apache Kafka producers
  • describe Apache Kafka consumers
  • describe Apache Kafka brokers
  • describe common hardware and OS specifications and their impact in Apache Kafka
  • describe the main options to deploy Apache Kafka
  • deploy Apache Kafka to Red Hat and CentOS
  • deploy Apache Kafka to Puppet
  • add and remove a broker in Apache Kafka
  • move data and partitions in Apache Kafka for performance purposes
  • add a new topic in Apache Kafka
  • scale a producer in Apache Kafka
  • scale a consumer in Apache Kafka
  • monitor Apache Kafka using the web console
  • monitor Apache Kafka using the offset monitor
  • monitor Apache Kafka using Graphite
  • monitor Apache Kafka using JMX
  • monitor Apache Kafka using the log files
  • tune the Linux kernel for Apache Kafka
  • tune Linux systems disk throughput for Apache Kafka
  • tune the Java VM for Apache Kafka
  • configure and manage Apache Kafka

Overview/Description
Apache Kafka's unique architecture enables huge scalability, but it must be deployed and managed in a considered fashion. This course covers the basic concepts of Apache Kafka, and considerations for deploying Kafka and managing servers.

Target Audience
Developers, IT Operations engineers, and DevOPs engineers looking to implement and manage Apache Kafka

Clustering

Course Number:
df_apka_a05_it_enus
Lesson Objectives

Clustering

  • start the course
  • describe the availability and durability guarantees of Apache Kafka
  • set up and configure a multi-broker cluster
  • configure and balance logs and leadership for replication
  • describe how log compaction works
  • configure and use the log cleaner
  • increase the replication factor for a topic
  • add servers to a Kafka cluster and migrate data to the new servers
  • mirror data between clusters such as between two data centers
  • balance replicas across zones or racks |w to prevent rack-failure data loss
  • set preferred replicas for leadership and handle unclean leader elections

Overview/Description
What makes Apache Kafka so powerful and fault-tolerant is its clustering capabilities. In this course, you will learn how to create and manage clusters in Kafka.

Target Audience
Developers, IT Operations engineers, and DevOPs engineers looking to implement and manage Apache Kafka

Kafka Integration with Spark

Course Number:
df_apka_a03_it_enus
Lesson Objectives

Kafka Integration with Spark

  • start the course
  • install and configure the Spark Streaming package for Kafka
  • read data into Spark from Kafka
  • read data in parallel into Spark from Kafka
  • write data back to Kafka from Spark
  • write data back to Kafka from Spark in parallel
  • create a direct stream to access Kafka data from Spark
  • use LocationStrategies and ConsumerStrategies to improve performance
  • use an RDD in cases where batch processing would be a better solution
  • use offsets to handle exactly-once semantics
  • use Kafka and Spark to split words from sentences

Overview/Description
Apache Kafka can easily integrate with Apache Spark to allow processing of the data entered into Kafka. In this course, you will discover how to integrate Kafka with Spark.

Target Audience
Developers, IT Operations engineers, and DevOPs engineers looking to implement and manage Apache Kafka

Kafka Integration with Storm

Course Number:
df_apka_a04_it_enus
Lesson Objectives

Kafka Integration with Storm

  • start the course
  • install and configure Storm and Kafka
  • describe the Storm-Kafka pipeline for processing events
  • configure the Kafka Spout for adding stream data into Storm
  • read data from Kafka into Storm
  • write data back to Kafka from a Storm Bolt
  • create a simple Kafka-Storm wordcount application
  • use the Kafka-Storm-Starter to bootstrap creating Kafka-Storm applications
  • describe how Kafka can be integrated with Spark and Storm

Overview/Description
Apache Kafka can easily integrate with Apache Storm to allow processing of the data entered into Kafka. In this course, you will learn how to integrate Kafka with Storm.

Target Audience
Developers, IT Operations engineers, and DevOPs engineers looking to implement and manage Apache Kafka

Real-time Applications

Course Number:
df_apka_a06_it_enus
Lesson Objectives

Real-time Applications

  • start the course
  • describe the real-time capabilities of Kafka
  • install and set up the Twitter Streaming API and framework to create a real-time Twitter application
  • create a Kafka Producer to process tweets
  • create a Consumer to transfer tweets to Spark
  • use Spark to process tweets and take action on specific words
  • create a Consumer to transfer tweets to Storm
  • use Storm to process the words in incoming tweets
  • use Kafka, Spark, and Storm in a pipeline to process words in a sentence
  • create a simple real-time application using Kafka, Spark, and Storm

Overview/Description
A major feature of Apache Kafka is building real-time applications that react to data streams. In this course, you will discover how to create real-time applications in Kafka.

Target Audience
Developers, IT Operations engineers, and DevOPs engineers looking to implement and manage Apache Kafka

Close Chat Live